Vladivostok
- Oceania > Australia > New South Wales > Sydney (0.14)
- Europe > Russia (0.04)
- Asia > Russia > Far Eastern Federal District > Primorsky Krai > Vladivostok (0.04)
- (2 more...)
Optimized Machine Learning Methods for Studying the Thermodynamic Behavior of Complex Spin Systems
Kapitan, Dmitrii, Ovchinnikov, Pavel, Soldatov, Konstantin, Andriushchenko, Petr, Kapitan, Vitalii
This paper presents a systematic study of the application of convolutional neural networks (CNNs) as an efficient and versatile tool for the analysis of critical and low-temperature phase states in spin system models. The problem of calculating the dependence of the average energy on the spatial distribution of exchange integrals for the Edwards-Anderson model on a square lattice with frustrated interactions is considered. We further construct a single convolutional classifier of phase states of the ferromagnetic Ising model on square, triangular, honeycomb, and kagome lattices, trained on configurations generated by the Swendsen-Wang cluster algorithm. Computed temperature profiles of the averaged posterior probability of the high-temperature phase form clear S-shaped curves that intersect in the vicinity of the theoretical critical temperatures and allow one to determine the critical temperature for the kagome lattice without additional retraining. It is shown that convolutional models substantially reduce the root-mean-square error (RMSE) compared with fully connected architectures and efficiently capture complex correlations between thermodynamic characteristics and the structure of magnetic correlated systems.
- Asia > Singapore (0.05)
- North America > United States > California > Santa Clara County > Mountain View (0.04)
- Europe > Russia (0.04)
- (3 more...)
- Oceania > Australia > New South Wales > Sydney (0.14)
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.04)
- Asia > Russia > Far Eastern Federal District > Primorsky Krai > Vladivostok (0.04)
- (3 more...)
Identifying internal patterns in (1+1)-dimensional directed percolation using neural networks
Parkhomenko, Danil, Ovchinnikov, Pavel, Soldatov, Konstantin, Kapitan, Vitalii, Chitov, Gennady Y.
In this paper we present a neural network-based method for the automatic detection of phase transitions and classification of hidden percolation patterns in a (1+1)-dimensional replication process. The proposed network model is based on the combination of CNN, TCN and GRU networks, which are trained directly on raw configurations without any manual feature extraction. The network reproduces the phase diagram and assigns phase labels to configurations. It shows that deep architectures are capable of extracting hierarchical structures from the raw data of numerical experiments.
- Europe > Russia (0.05)
- Asia > Singapore (0.05)
- Asia > Russia > Far Eastern Federal District > Primorsky Krai > Vladivostok (0.04)
- (2 more...)
OptimalThinkingBench: Evaluating Over and Underthinking in LLMs
Aggarwal, Pranjal, Kim, Seungone, Lanchantin, Jack, Welleck, Sean, Weston, Jason, Kulikov, Ilia, Saha, Swarnadeep
Thinking LLMs solve complex tasks at the expense of increased compute and overthinking on simpler problems, while non-thinking LLMs are faster and cheaper but underthink on harder reasoning problems. This has led to the development of separate thinking and non-thinking LLM variants, leaving the onus of selecting the optimal model for each query on the end user. We introduce OptimalThinkingBench, a unified benchmark that jointly evaluates overthinking and underthinking in LLMs and also encourages the development of optimally-thinking models that balance performance and efficiency. Our benchmark comprises two sub-benchmarks: OverthinkingBench, featuring simple math and general queries in 72 domains, and UnderthinkingBench, containing 11 challenging reasoning tasks along with harder math problems. Using novel thinking-adjusted accuracy metrics, we extensively evaluate 33 different thinking and non-thinking models and show that no model is able to optimally think on our benchmark. Thinking models often overthink for hundreds of tokens on the simplest user queries without improving performance. In contrast, large non-thinking models underthink, often falling short of much smaller thinking models. We further explore several methods to encourage optimal thinking, but find that these approaches often improve on one sub-benchmark at the expense of the other, highlighting the need for better unified and optimal models in the future.
- Europe > Russia > Northwestern Federal District > Kaliningrad Oblast > Kaliningrad (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- North America > United States (0.04)
- (9 more...)
- Leisure & Entertainment (1.00)
- Health & Medicine (1.00)
- Media > Music (0.94)
- Education (0.68)
MMeViT: Multi-Modal ensemble ViT for Post-Stroke Rehabilitation Action Recognition
Kim, Ye-eun, Lim, Suhyeon, Choi, Andrew J.
Rehabilitation therapy for stroke patients faces a supply shortage despite the increasing demand. To address this issue, remote monitoring systems that reduce the burden on medical staff are emerging as a viable alternative. A key component of these remote monitoring systems is Human Action Recognition (HAR) technology, which classifies actions. However, existing HAR studies have primarily focused on non-disable individuals, making them unsuitable for recognizing the actions of stroke patients. HAR research for stroke has largely concentrated on classifying relatively simple actions using machine learning rather than deep learning. In this study, we designed a system to monitor the actions of stroke patients, focusing on domiciliary upper limb Activities of Daily Living (ADL). Our system utilizes IMU (Inertial Measurement Unit) sensors and an RGB-D camera, which are the most common modalities in HAR. We directly collected a dataset through this system, investigated an appropriate preprocess and proposed a deep learning model suitable for processing multimodal data. We analyzed the collected dataset and found that the action data of stroke patients is less clustering than that of non-disabled individuals. Simultaneously, we found that the proposed model learns similar tendencies for each label in data with features that are difficult to clustering. This study suggests the possibility of expanding the deep learning model, which has learned the action features of stroke patients, to not only simple action recognition but also feedback such as assessment contributing to domiciliary rehabilitation in future research. The code presented in this study is available at https://github.com/ye-Kim/MMeViT.
- North America > United States > Missouri > Jackson County > Kansas City (0.14)
- North America > United States > Texas > Travis County > Austin (0.04)
- North America > United States > California > San Diego County > San Diego (0.04)
- (6 more...)
- Research Report > New Finding (0.55)
- Research Report > Experimental Study (0.34)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Health & Medicine > Therapeutic Area > Hematology (1.00)
- Health & Medicine > Therapeutic Area > Cardiology/Vascular Diseases (1.00)
LongProc: Benchmarking Long-Context Language Models on Long Procedural Generation
Ye, Xi, Yin, Fangcong, He, Yinghui, Zhang, Joie, Yen, Howard, Gao, Tianyu, Durrett, Greg, Chen, Danqi
Existing benchmarks for evaluating long-context language models (LCLMs) primarily focus on long-context recall, requiring models to produce short responses based on a few critical snippets while processing thousands of irrelevant tokens. We introduce LongProc (Long Procedural Generation), a new benchmark that requires both the integration of highly dispersed information and long-form generation. LongProc consists of six diverse procedural generation tasks, such as extracting structured information from HTML pages into a TSV format and executing complex search procedures to create travel plans. These tasks challenge LCLMs by testing their ability to follow detailed procedural instructions, synthesize and reason over dispersed information, and generate structured, long-form outputs (up to 8K tokens). Furthermore, as these tasks adhere to deterministic procedures and yield structured outputs, they enable reliable rule-based evaluation. We evaluate 17 LCLMs on LongProc across three difficulty levels, with maximum numbers of output tokens set at 500, 2K, and 8K. Notably, while all tested models claim a context window size above 32K tokens, open-weight models typically falter on 2K-token tasks, and closed-source models like GPT-4o show significant degradation on 8K-token tasks. Further analysis reveals that LCLMs struggle to maintain long-range coherence in long-form generations. These findings highlight critical limitations in current LCLMs and suggest substantial room for improvement. Data and code available at: https://princeton-pli.github.io/LongProc
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.05)
- Europe > Sweden > Stockholm > Stockholm (0.05)
- Europe > Germany > Baden-Württemberg > Stuttgart Region > Stuttgart (0.05)
- (33 more...)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science > Problem Solving (0.92)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.91)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.69)
North Korean troops in Ukraine 'fair game', US warns Russia as war rages on
United States defence secretary Lloyd Austin has waded in on reports that North Korea was preparing to enter the Ukraine war with troops. "If they are co-belligerents, if their intention is to participate in this war on Russia's behalf, that is a very, very serious issue," Austin said. Austin was returning from his fourth visit to Kyiv, where he announced a 400m package of US weapons for Ukraine. John Kirby, White House national security spokesman, said Washington believes that at least 3,000 North Korean soldiers arrived this month by sea to Vladivostok, Russia's largest Pacific port. "These soldiers then travelled onward to multiple Russian military training sites in eastern Russia, where they are currently undergoing training," Kirby said on Wednesday.
- Africa (0.30)
- Europe > Ukraine > Kyiv Oblast > Kyiv (0.26)
- Asia > Russia > Far Eastern Federal District > Primorsky Krai > Vladivostok (0.26)
- (11 more...)
- Government > Military (1.00)
- Government > Regional Government > Europe Government > Russia Government (0.93)
- Government > Regional Government > Asia Government > Russia Government (0.93)
A Russian Jeopardy! Data Set for Question-Answering Systems
Question answering (QA) is one of the most common NLP tasks that relates to named entity recognition, fact extraction, semantic search and some other fields. In industry, it is much appreciated in chatbots and corporate information systems. It is also a challenging task that attracted the attention of a very general audience at the quiz show Jeopardy! In this article we describe a Jeopardy!-like Russian QA data set collected from the official Russian quiz database Chgk (che ge ka). The data set includes 379,284 quiz-like questions with 29,375 from the Russian analogue of Jeopardy! - "Own Game". We observe its linguistic features and the related QA-task. We conclude about perspectives of a QA competition based on the data set collected from this database.
- Europe > Russia (0.14)
- Asia > Russia > Ural Federal District > Tyumen Oblast > Tyumen (0.05)
- Europe > United Kingdom > England (0.04)
- (8 more...)
Space warfare: US, China, and Russia are gearing up for the next frontier of armed conflict
Arthel Neville welcomes former U.S. Defense Intelligence Officer Rebekah Koffler to discuss the massive global cyberattack that had impacted several federal agencies. The next big war may be fought in space. As the Pentagon is gearing up for a future celestial conflict, so are our chief adversaries, China and Russia. Here's why "Star Wars" is no longer merely a topic of science fiction. The best way to avoid space warfare is to be ready for it. On Dec. 28, Elon Musk's Space X launched into space the Pentagon's highly secretive X-37B Orbital Test Vehicle, an unmanned reusable robotic spacecraft operated by the Air Force, in collaboration with Space Force.
- Europe > Ukraine (0.05)
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.05)
- Asia > Middle East > Iran (0.05)
- (19 more...)
- Government > Regional Government > North America Government > United States Government (1.00)
- Government > Military (1.00)
- Government > Regional Government > Asia Government (0.72)